Goto

Collaborating Authors

 stable material


WyckoffDiff - A Generative Diffusion Model for Crystal Symmetry

Kelvinius, Filip Ekström, Andersson, Oskar B., Parackal, Abhijith S., Qian, Dong, Armiento, Rickard, Lindsten, Fredrik

arXiv.org Artificial Intelligence

Crystalline materials often exhibit a high level of symmetry. However, most generative models do not account for symmetry, but rather model each atom without any constraints on its position or element. We propose a generative model, Wyckoff Diffusion (WyckoffDiff), which generates symmetry-based descriptions of crystals. This is enabled by considering a crystal structure representation that encodes all symmetry, and we design a novel neural network architecture which enables using this representation inside a discrete generative model framework. In addition to respecting symmetry by construction, the discrete nature of our model enables fast generation. We additionally present a new metric, Fr\'echet Wrenformer Distance, which captures the symmetry aspects of the materials generated, and we benchmark WyckoffDiff against recently proposed generative models for crystal generation.


Establishing baselines for generative discovery of inorganic crystals

Szymanski, Nathan J., Bartel, Christopher J.

arXiv.org Artificial Intelligence

Generative artificial intelligence offers a promising avenue for materials discovery, yet its advantages over traditional methods remain unclear. In this work, we introduce and benchmark two baseline approaches - random enumeration of charge-balanced prototypes and data-driven ion exchange of known compounds - against three generative models: a variational autoencoder, a large language model, and a diffusion model. Our results show that established methods such as ion exchange perform comparably well in generating stable materials, although many of these materials tend to closely resemble known compounds. In contrast, generative models excel at proposing novel structural frameworks and, when sufficient training data is available, can more effectively target properties such as electronic band gap and bulk modulus while maintaining a high stability rate. To enhance the performance of both the baseline and generative approaches, we implement a post-generation screening step in which all proposed structures are passed through stability and property filters from pre-trained machine learning models including universal interatomic potentials. This low-cost filtering step leads to substantial improvement in the success rates of all methods, remains computationally efficient, and ultimately provides a practical pathway toward more effective generative strategies for materials discovery.


FlowLLM: Flow Matching for Material Generation with Large Language Models as Base Distributions

Sriram, Anuroop, Miller, Benjamin Kurt, Chen, Ricky T. Q., Wood, Brandon M.

arXiv.org Machine Learning

Material discovery is a critical area of research with the potential to revolutionize various fields, including carbon capture, renewable energy, and electronics. However, the immense scale of the chemical space makes it challenging to explore all possible materials experimentally. In this paper, we introduce FlowLLM, a novel generative model that combines large language models (LLMs) and Riemannian flow matching (RFM) to design novel crystalline materials. FlowLLM first fine-tunes an LLM to learn an effective base distribution of meta-stable crystals in a text representation. After converting to a graph representation, the RFM model takes samples from the LLM and iteratively refines the coordinates and lattice parameters. Our approach significantly outperforms state-of-the-art methods, increasing the generation rate of stable materials by over three times and increasing the rate for stable, unique, and novel crystals by $\sim50\%$ - a huge improvement on a difficult problem. Additionally, the crystals generated by FlowLLM are much closer to their relaxed state when compared with another leading model, significantly reducing post-hoc computational cost.


Fine-Tuned Language Models Generate Stable Inorganic Materials as Text

Gruver, Nate, Sriram, Anuroop, Madotto, Andrea, Wilson, Andrew Gordon, Zitnick, C. Lawrence, Ulissi, Zachary

arXiv.org Artificial Intelligence

We propose fine-tuning large language models for generation of stable materials. While unorthodox, fine-tuning large language models on text-encoded atomistic data is simple to implement yet reliable, with around 90% of sampled structures obeying physical constraints on atom positions and charges. Using energy above hull calculations from both learned ML potentials and gold-standard DFT calculations, we show that our strongest model (fine-tuned LLaMA-2 70B) can generate materials predicted to be metastable at about twice the rate (49% vs 28%) of CD-VAE, a competing diffusion model. Because of text prompting's inherent flexibility, our models can simultaneously be used for unconditional generation of stable material, infilling of partial structures and text-conditional generation. Finally, we show that language models' ability to capture key symmetries of crystal structures improves with model scale, suggesting that the biases of pretrained LLMs are surprisingly well-suited for atomistic data. Large language models (LLMs) are trained to compress large text datasets, but can also act as strong foundations for non-text data (Delétang et al., 2023). As compressors, LLMs extract common patterns and find simple programs that can produce them (Goldblum et al., 2023; Sutskever, 2023), regardless of the data's origin. Alongside generality, LLM pre-training also gives rise to sample efficiency, as in-context learning and fine-tuning require far fewer training examples to identify salient patterns than training a model from scratch (Brown et al., 2020). The generality and sample efficiency of LLMs make them particular promising for scientific problems, where data are often limited, collected from diverse sources, or challenging for non-experts to interpret. In materials science, for example, the number of known stable materials is relatively small, and the data describing each material are diverse, including composition, structure, and complex properties. LLMs can learn generalizable rules from a small number of examples (Zhu et al., 2023), combine modalities into a single model (Moon et al., 2023), and provide users with a text-based interface. A text interface, in particular, has the potential to improve access to scientific discovery (White, 2023); LLMs can use text to describe new observations, or, in design applications (e.g. In this work, we show that fine-tuned LLMs can generate the three-dimensional structure of stable crystals as text (Figure 1).


Google DeepMind AI Breakthrough Could Help Battery and Chip Development

TIME - Tech

Researchers at Google DeepMind have used artificial intelligence to predict the structures of more than 2 million new materials, in a breakthrough that could have wide-reaching benefits in sectors such as renewable energy and computing. DeepMind published 381,000 of the 2.2 million crystal structures that it predicts to be most stable. The breakthrough increases the number of known stable materials by a factor of ten. Although the materials will still need to be synthesized and tested, steps which can take months or even years, the latest development is expected to accelerate the discovery of new materials, which will be required for applications such as energy storage, solar cells, and superconductor chips. "While materials play a very critical role in almost any technology, we as humanity know only about a few tens of thousands of stable materials," says Ekin Dogus Cubuk, a Staff Research Scientist at Google Brain, who worked on the DeepMind AI tool, known as Graph Networks for Materials Exploration (GNoME).


Google DeepMind's new AI tool helped create more than 700 new materials

MIT Technology Review

GNoME can be described as AlphaFold for materials discovery, according to Ju Li, a materials science and engineering professor at the Massachusetts Institute of Technology. AlphaFold, a DeepMind AI system announced in 2020, predicts the structures of proteins with high accuracy and has since advanced biological research and drug discovery. Thanks to GNoME, the number of known stable materials has grown almost tenfold, to 421,000. "While materials play a very critical role in almost any technology, we as humanity know only a few tens of thousands of stable materials," said Dogus Cubuk, materials discovery lead at Google DeepMind, at a press briefing. To discover new materials, scientists combine elements across the periodic table.


Physics Guided Deep Learning for Generative Design of Crystal Materials with Symmetry Constraints

Zhao, Yong, Siriwardane, Edirisuriya M. Dilanga, Wu, Zhenyao, Fu, Nihang, Al-Fahdi, Mohammed, Hu, Ming, Hu, Jianjun

arXiv.org Artificial Intelligence

Discovering new materials is a challenging task in materials science crucial to the progress of human society. Conventional approaches based on experiments and simulations are labor-intensive or costly with success heavily depending on experts' heuristic knowledge. Here, we propose a deep learning based Physics Guided Crystal Generative Model (PGCGM) for efficient crystal material design with high structural diversity and symmetry. Our model increases the generation validity by more than 700\% compared to FTCP, one of the latest structure generators and by more than 45\% compared to our previous CubicGAN model. Density Functional Theory (DFT) calculations are used to validate the generated structures with 1,869 materials out of 2,000 are successfully optimized and deposited into the Carolina Materials Database \url{www.carolinamatdb.org}, of which 39.6\% have negative formation energy and 5.3\% have energy-above-hull less than 0.25 eV/atom, indicating their thermodynamic stability and potential synthesizability.


Using AI to tackle the challenge of materials structure prediction

AIHub

Image reproduced under a CC BY-NC 4.0 licence. Researchers have designed a machine learning method that can predict the structure of new materials. The researchers, from Cambridge and Linköping Universities, have designed a way to predict the structure of materials given its constitutive elements. The results are reported in the journal Science Advances. The arrangement of atoms in a material determines its properties.

  Country: Europe > Sweden > Östergötland County > Linköping (0.26)
  Industry: Energy (0.33)

Data-driven discovery of novel 2D materials by deep generative models

Lyngby, Peder, Thygesen, Kristian Sommer

arXiv.org Artificial Intelligence

Efficient algorithms to generate candidate crystal structures with good stability properties can play a key role in data-driven materials discovery. Here we show that a crystal diffusion variational autoencoder (CDVAE) is capable of generating two-dimensional (2D) materials of high chemical and structural diversity and formation energies mirroring the training structures. Specifically, we train the CDVAE on 2615 2D materials with energy above the convex hull $\Delta H_{\mathrm{hull}}< 0.3$ eV/atom, and generate 5003 materials that we relax using density functional theory (DFT). We also generate 14192 new crystals by systematic element substitution of the training structures. We find that the generative model and lattice decoration approach are complementary and yield materials with similar stability properties but very different crystal structures and chemical compositions. In total we find 11630 predicted new 2D materials, where 8599 of these have $\Delta H_{\mathrm{hull}}< 0.3$ eV/atom as the seed structures, while 2004 are within 50 meV of the convex hull and could potentially be synthesized. The relaxed atomic structures of all the materials are available in the open Computational 2D Materials Database (C2DB). Our work establishes the CDVAE as an efficient and reliable crystal generation machine, and significantly expands the space of 2D materials.